152 research outputs found

    What if Achilles and the tortoise were to bargain? An argument against interim agreements

    Get PDF
    Zeno's paradoxes of motion, which claim that moving from one point to another cannot be accomplished in finite time, seem to be of serious concern when moving towards an agreement is concerned. Parkinson's Law of Triviality implies that such an agreement cannot be reached in finite time. By explicitly modeling dynamic processes of reaching interim agreements and using arguments similar to Zeno's, we show that if utilities are von Neumann-Morgenstern, then no such process can bring about an agreement in finite time in linear bargaining problems. To extend this result for all bargaining problems, we characterize a particular path illustrated by \cite{ra}, and show that no agreement is reached along this path in finite time.Zeno's paradox, bargaining problems, interim agreements, vNM utility

    S5 Knowledge Without Partitions

    Get PDF

    Bayesianism without Learning

    Get PDF
    According to the standard definition, a Bayesian agent is one who forms his posterior belief by conditioning his prior belief on what he has learned, that is, on facts of which he has become certain. Here it is shown that Bayesianism can be described without assuming that the agent acquires any certain information; an agent is Bayesian if his prior, when conditioned on his posterior belief, agrees with the latter. This condition is shown to characterize Bayesian models.Bayesian updating, prior and posterior

    Learning to Play Games in Extensive Form by Valuation

    Full text link
    A valuation for a player in a game in extensive form is an assignment of numeric values to the players moves. The valuation reflects the desirability moves. We assume a myopic player, who chooses a move with the highest valuation. Valuations can also be revised, and hopefully improved, after each play of the game. Here, a very simple valuation revision is considered, in which the moves made in a play are assigned the payoff obtained in the play. We show that by adopting such a learning process a player who has a winning strategy in a win-lose game can almost surely guarantee a win in a repeated game. When a player has more than two payoffs, a more elaborate learning procedure is required. We consider one that associates with each move the average payoff in the rounds in which this move was made. When all players adopt this learning procedure, with some perturbations, then, with probability 1, strategies that are close to subgame perfect equilibrium are played after some time. A single player who adopts this procedure can guarantee only her individually rational payoff

    One Observation behind Two-Envelope Puzzles

    Get PDF
    In two famous and popular puzzles a participant is required to compare two numbers of which she is shown only one. In the first one there are two envelopes with money in them. The sum of money in one of the envelopes is twice as large as the other sum. An envelope is selected at random and handed to you. If the sum in this envelope is x, then the sum in the other one is (1/2)(2x) + (1/2)(0.5x) = 1.25x. Hence, you are better off switching to the other envelope no matter what sum you see, which is paradoxical. In the second puzzle two distinct numbers are written on two slips of paper. One of them is selected at random and you observe it. How can you guess, with probability greater than 1/2 of being correct, whether this number is the larger or the smaller? We show that there is one principle behind the two puzzles: The ranking of n random variables X1, ... , Xn cannot be independent of each of them, unless the ranking is fixed. Thus, unless there is nothing to be learned about the ranking, there must be at least one variable the observation of which conveys information about it.two envelope paradox

    Learning to play games in extensive form by valuation

    Get PDF
    A valuation for a board game is an assignment of numeric values to different states of the board. The valuation reflects the desirability of the states for the player. It can be used by a player to decide on her next move during the play. We assume a myopic player, who chooses a move with the highest valuation. Valuations can also be revised, and hopefully improved, after each play of the game. Here, a very simple valuation revision is considered, in which the states of the board visited in a play are assigned the payoff obtained in the play. We show that by adopting such a learning process a player who has a winning strategy in a win-lose game can almost surely guarantee a win in a repeated game. When a player has more than two payoffs, a more elaborate learning procedure is required. We consider one that associates with each state the average payoff in the rounds in which this node was reached. When all players adopt this learning procedure, with some perturbations, then, with probability 1, strategies that are close to subgame perfect equilibrium are played after some time. A single player who adopts this procedure can guarantee only her individually rational payoff.reinforcement learning

    Between Liberalism and Democracy

    Get PDF
    We study a class of voting rules that bridge between majoritarianism and liberalism. An outcome of the vote specifies who among the voters are eligible to a certain right or qualification. Each outcome serves also as a permissible ballot. We characterize axiomatically a family of rules parameterized by the weight each individual has in determining his or her qualification. In one extreme case, the Liberal Rule, each individual's qualification is determined by her. In the other, an individual's qualification is determined by a majority. We also propose a formalization of self-determination, and apply it in a characterization of the Liberal Rule.liberalism, voting rules, social choice
    • ā€¦
    corecore